Goto

Collaborating Authors

 compute-optimal neural scaling law


4+3 Phases of Compute-Optimal Neural Scaling Laws

Neural Information Processing Systems

We consider the solvable neural scaling model with three parameters: data complexity, target complexity, and model-parameter-count. We use this neural scaling model to derive new predictions about the compute-limited, infinite-data scaling law regime. To train the neural scaling model, we run one-pass stochastic gradient descent on a mean-squared loss. We derive a representation of the loss curves which holds over all iteration counts and improves in accuracy as the model parameter count grows. The phase boundaries are determined by the relative importance of model capacity, optimizer noise, and embedding of the features.